Refactor interally to RawBatch and CleanBatch wrapper types
#57
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
I was adding tests for
parse_stac_ndjson_to_arrowwhen I discovered that that was inadvertently broken in #53 due to a refactor in the schema inference. It has been confusing to work with untypedpyarrow.RecordBatchclasses because they're a black box, and we have two distinct data schemas we're working with: one that is as close to the raw JSON as possible, and another that conforms to our STAC GeoParquet schema.This PR refactors this internally by adding
RawBatchandCleanBatchwrapper types (open to better naming suggestions, but these are not public, so we can easily rename in the future). They both hold an internalpyarrow.RecordBatchbutRawBatchaligns as much as possible to the raw STAC JSON representation, whileCleanBatchaligns to the STAC-GeoParquet schema.These wrapper types make it much easier to reason about the shape of the data at different points of the pipeline.
Change list
RawBatchandCleanBatchfor more reliable internal typing